新兴智能反射表面(IRS)技术介绍了可见光通信(VLC)系统中受控光传播的电位。这一概念为新应用程序打开了门,其中可以改变通道本身以实现特定的关键性能指标。在本文中,在开放文献中首次调查IRSS可以在提高采用非正交多址(NOMA)的VLC系统中的链路可靠性方面的作用。我们为NOMA和IRS参数的联合优化提出了一个框架,并表明它在链路可靠性方面提供了显着的增强。当VLC通道受阻堵塞和随机设备方向时,增强更加明显。
translated by 谷歌翻译
Over the years, sequential Monte Carlo (SMC) and, equivalently, particle filter (PF) theory has gained substantial attention from researchers. However, the performance of the resampling methodology, also known as offspring selection, has not advanced recently. We propose two deterministic offspring selection methods, which strive to minimize the Kullback-Leibler (KL) divergence and the total variation (TV) distance, respectively, between the particle distribution prior and subsequent to the offspring selection. By reducing the statistical distance between the selected offspring and the joint distribution, we obtain a heuristic search procedure that performs superior to a maximum likelihood search in precisely those contexts where the latter performs better than an SMC. For SMC and particle Markov chain Monte Carlo (pMCMC), our proposed offspring selection methods always outperform or compare favorably with the two state-of-the-art resampling schemes on two models commonly used as benchmarks from the literature.
translated by 谷歌翻译
Topic modeling is widely used for analytically evaluating large collections of textual data. One of the most popular topic techniques is Latent Dirichlet Allocation (LDA), which is flexible and adaptive, but not optimal for e.g. short texts from various domains. We explore how the state-of-the-art BERTopic algorithm performs on short multi-domain text and find that it generalizes better than LDA in terms of topic coherence and diversity. We further analyze the performance of the HDBSCAN clustering algorithm utilized by BERTopic and find that it classifies a majority of the documents as outliers. This crucial, yet overseen problem excludes too many documents from further analysis. When we replace HDBSCAN with k-Means, we achieve similar performance, but without outliers.
translated by 谷歌翻译
Multi-agent artificial intelligence research promises a path to develop intelligent technologies that are more human-like and more human-compatible than those produced by "solipsistic" approaches, which do not consider interactions between agents. Melting Pot is a research tool developed to facilitate work on multi-agent artificial intelligence, and provides an evaluation protocol that measures generalization to novel social partners in a set of canonical test scenarios. Each scenario pairs a physical environment (a "substrate") with a reference set of co-players (a "background population"), to create a social situation with substantial interdependence between the individuals involved. For instance, some scenarios were inspired by institutional-economics-based accounts of natural resource management and public-good-provision dilemmas. Others were inspired by considerations from evolutionary biology, game theory, and artificial life. Melting Pot aims to cover a maximally diverse set of interdependencies and incentives. It includes the commonly-studied extreme cases of perfectly-competitive (zero-sum) motivations and perfectly-cooperative (shared-reward) motivations, but does not stop with them. As in real-life, a clear majority of scenarios in Melting Pot have mixed incentives. They are neither purely competitive nor purely cooperative and thus demand successful agents be able to navigate the resulting ambiguity. Here we describe Melting Pot 2.0, which revises and expands on Melting Pot. We also introduce support for scenarios with asymmetric roles, and explain how to integrate them into the evaluation protocol. This report also contains: (1) details of all substrates and scenarios; (2) a complete description of all baseline algorithms and results. Our intention is for it to serve as a reference for researchers using Melting Pot 2.0.
translated by 谷歌翻译
Network-based analyses of dynamical systems have become increasingly popular in climate science. Here we address network construction from a statistical perspective and highlight the often ignored fact that the calculated correlation values are only empirical estimates. To measure spurious behaviour as deviation from a ground truth network, we simulate time-dependent isotropic random fields on the sphere and apply common network construction techniques. We find several ways in which the uncertainty stemming from the estimation procedure has major impact on network characteristics. When the data has locally coherent correlation structure, spurious link bundle teleconnections and spurious high-degree clusters have to be expected. Anisotropic estimation variance can also induce severe biases into empirical networks. We validate our findings with ERA5 reanalysis data. Moreover we explain why commonly applied resampling procedures are inappropriate for significance evaluation and propose a statistically more meaningful ensemble construction framework. By communicating which difficulties arise in estimation from scarce data and by presenting which design decisions increase robustness, we hope to contribute to more reliable climate network construction in the future.
translated by 谷歌翻译
我们提出了对标准重新结构体系结构的简单修改 - 在特征空间上进行L2正则化 - 从前提出的深层确定性不确定性(DDU)基准中,它显着改善了分布外(OOD)的性能。这种变化还引起了早期神经塌陷(NC),我们证明这是一种更有可能的OOD性能的效果。我们的方法在基准的一小部分训练时间中实现了可比或优质的OOD检测分数和分类精度。此外,它基本上改善了多个随机初始化模型的最坏情况。尽管我们不建议NC是深神经网络(DNN)中OOD行为的唯一机制或全面解释,但我们认为NC的简单数学和几何结构可以为对未来工作中这种复杂现象的分析提供一个框架。
translated by 谷歌翻译
语言的视觉基础旨在用多种视觉知识来源(例如图像和视频)丰富语言表示。尽管视觉接地是一个深入研究的领域,但视觉接地的语言方面并没有得到太多关注。本研究调查了单词嵌入的语法视觉基础。我们在两个视觉和语言空间之间提出了一种隐式对齐技术,其中语言之间的文本信息相互作用以丰富预训练的文本单词嵌入。我们专注于实验中的三种语言,即英语,阿拉伯语和德语。我们获得了这些语言的视觉接地矢量表示形式,并研究了一种或多种语言的视觉接地是否改善了嵌入在单词相似性和分类基准上的嵌入性能。我们的实验表明,语法知识可以改善类似语言(例如德语和英语)的扎根嵌入性能。但是,德语或英语用阿拉伯语的语言基础导致单词相似性基准的性能略有降解。另一方面,我们观察到了分类基准的相反趋势,而阿拉伯语对英语的进步最大。在讨论部分中,提出了这些发现的几个原因。我们希望我们的实验为进一步研究的基线提供了有关语法间视觉接地的基准。
translated by 谷歌翻译
深度代码生成是软件工程深度学习(DL4SE)的主题,该主题采用神经模型来为预期功能生成代码。由于端到端的神经方法缺乏对域知识和软件层次结构的认识,因此结果通常需要手动校正。为了系统地探索代码生成的潜在改进,我们让IT参与从意图到实现的整个自上而下的发展,这在有限的范围中是可能的。在此过程中,它受益于大量样本,功能和知识。作为基金会,我们建议对代码数据(即代码分类法)建立分类法,利用代码信息的分类。此外,我们引入了三层语义金字塔(SP)以关联文本数据和代码数据。它标识了不同的抽象水平的信息,因此介绍了有关开发的领域知识,并揭示了软件的层次结构。此外,我们提出了一个语义金字塔框架(SPF)作为方法,重点是高模块化和低复杂性的软件。 SPF将代码生成过程分为阶段,并为潜在的相互作用提供储量。最终,我们为SPF构思了应用程序范围。
translated by 谷歌翻译
知识图(kgs)中的实体类型信息(例如DBPEDIA,FREEBASE等)通常由于自动产生或人类策划而通常不完整。实体键入是在kg中分配或推断实体的语义类型的任务。本文介绍了\ textit {grand {grand},这是一种实体键入的新方法,利用RDF2VEC中的不同图形步行策略以及文本实体描述。 RDF2VEC首先生成图形步行,然后使用语言模型来获取图中每个节点的嵌入。这项研究表明,步行生成策略和嵌入模型对实体打字任务的性能有重大影响。所提出的方法的表现优于基准数据集DBPedia和Figer在kgs中的实体和小颗粒类别的实体。结果表明,订单感知RDF2VEC变体的组合以及文本实体描述的上下文嵌入可实现最佳结果。
translated by 谷歌翻译
分销语义提供了研究形态学语义的新方法。这项研究的重点是名词奇异人的语义及其在英语中的复数变种变体。我们的目标是比较两个模型的多元化概念化。一个模型(FRACSS)提出,在预测来自单数语义的复数语义时,应考虑所有奇异对。另一个模型(CCA)认为,多元化的概念化主要取决于基本单词的语义类别。我们根据大量的美国英语语音与两个模型预测的语义矢量相一致的大量语料库中复数代币的语音信号的方式进行比较。采用了两项措施:表单与义映射的性能以及形式距离和含义距离之间的相关性。结果收敛于CCA的优质比对。我们的结果表明,基于用法的多元化方法,其中给定单词自己的语义社区的优先级优于理论,根据该理论,多元化被概念化为基于高级抽象的过程。我们看到,经常被认为是一个高度抽象的概念,[+复数]可以通过中级部分概括的家庭更好地捕获。
translated by 谷歌翻译